Misspecification in Inverse Reinforcement Learning
نویسندگان
چکیده
The aim of Inverse Reinforcement Learning (IRL) is to infer a reward function R from policy pi. To do this, we need model how pi relates R. In the current literature, most common models are optimality, Boltzmann rationality, and causal entropy maximisation. One primary motivations behind IRL human preferences behaviour. However, true relationship between behaviour much more complex than any currently used in IRL. This means that they misspecified, which raises worry might lead unsound inferences if applied real-world data. this paper, provide mathematical analysis robust different misspecification, answer precisely demonstrator may differ each standard before leads faulty about We also introduce framework for reasoning misspecification IRL, together with formal tools can be easily derive robustness new models.
منابع مشابه
Repeated Inverse Reinforcement Learning
We introduce a novel repeated Inverse Reinforcement Learning problem: the agent has to act on behalf of a human in a sequence of tasks and wishes to minimize the number of tasks that it surprises the human by acting suboptimally with respect to how the human would have acted. Each time the human is surprised, the agent is provided a demonstration of the desired behavior by the human. We formali...
متن کاملBayesian Inverse Reinforcement Learning
Inverse Reinforcement Learning (IRL) is the problem of learning the reward function underlying a Markov Decision Process given the dynamics of the system and the behaviour of an expert. IRL is motivated by situations where knowledge of the rewards is a goal by itself (as in preference elicitation) and by the task of apprenticeship learning (learning policies from an expert). In this paper we sh...
متن کاملCooperative Inverse Reinforcement Learning
For an autonomous system to be helpful to humans and to pose no unwarranted risks, it needs to align its values with those of the humans in its environment in such a way that its actions contribute to the maximization of value for the humans. We propose a formal definition of the value alignment problem as cooperative inverse reinforcement learning (CIRL). A CIRL problem is a cooperative, parti...
متن کاملInverse Reinforcement Learning
Prescribes action to take for each state Probability distribution over next states given current state and action Describes desirability of being in a state.
متن کاملInverse Reinforcement Learning
Recently researches on imitation learning have shown that Markov Decision Processes (MDPs) are a powerful way to characterize this problem. Inverse reinforcement learning tries to describe observed behavior by ascertaining a reward function (or respectively a cost function) by solving a Markov Decision Problem. This paper shows three different approaches to find an optimal policy which mimics o...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i12.26766